Mistral
MistralFrench AI startup offering efficient and cost-effective language models, including Mistral 7B and Mixtral.
Total Models: 17
mistral-medium-2505
mistral-medium-2505

Mistral Medium 3 is an advanced Large Language Model (LLM) with state-of-the-art reasoning, knowledge, coding and vision capabilities.

chat-completion
conversational
image-classification
mistral-small-2503
mistral-small-2503

Enhanced Mistral Small 3 with multimodal capabilities and a 128k context length.

chat-completion
Completions
Conversational
Image classification
Question answering
Ministral-3B
Ministral-3B

Ministral 3B is a state-of-the-art Small Language Model (SLM) optimized for edge computing and on-device applications. As it is designed for low-latency and compute-efficient inference, it it also the perfect model for standard GenAI applications that have

chat-completion
Codestral-2501
Codestral-2501

Codestral 25.01 by Mistral AI is designed for code generation, supporting 80+ programming languages, and optimized for tasks like code completion and fill-in-the-middle

chat-completion
mistralai-Mistral-7B-Instruct-v01
mistralai-Mistral-7B-Instruct-v01

Model Details The Mistral7BInstructv0.1 Large Language Model (LLM) is a instruct finetuned version of the Mistral7Bv0.1 generative text model using a variety of publicly available conversation datasets. For full details of this model

chat-completion
Mistral-small
Mistral-small

Mistral Small can be used on any language-based task that requires high efficiency and low latency.

chat-completion
mistral-ocr-2503
mistral-ocr-2503

Document conversion to markdown with interleaved images and text

image-to-text
mistralai-Mixtral-8x7B-Instruct-v01
mistralai-Mixtral-8x7B-Instruct-v01

The Mixtral8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mixtral8x7B outperforms Llama 2 70B on most benchmarks with 6x faster inference. Mixtral8x7Bv0.1 is a decoderonly model with 8 distinct groups or the "experts". At every layer, for every token,

chat-completion
mistralai-Mixtral-8x22B-Instruct-v0-1
mistralai-Mixtral-8x22B-Instruct-v0-1

The Mixtral8x22BInstructv0.1 Large Language Model (LLM) is an instruct finetuned version of the Mixtral8x22Bv0.1. Inference samples Inference type|Python sample (Notebook)|CLI with YAML |||| Real time|<a href="https://aka.ms/azu

chat-completion
mistralai-Mistral-7B-v01
mistralai-Mistral-7B-v01

Model Details The Mistral7Bv0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral7Bv0.1 outperforms Llama 2 13B on all benchmarks tested. For full details of this model please read paper and [release b

text-generation
Mistral-large-2407
Mistral-large-2407

Mistral Large (2407) is an advanced Large Language Model (LLM) with state-of-the-art reasoning, knowledge and coding capabilities.

chat-completion
mistralai-Mixtral-8x22B-v0-1
mistralai-Mixtral-8x22B-v0-1

The Mixtral8x22B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. Mixtral8x22Bv0.1 is a pretrained base model and therefore does not have any moderation mechanisms. Evaluation Results [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/Hugg

text-generation
Mistral-Nemo
Mistral-Nemo

Mistral Nemo is a cutting-edge Language Model (LLM) boasting state-of-the-art reasoning, world knowledge, and coding capabilities within its size category.

chat-completion
mistralai-Mistral-7B-Instruct-v0-2
mistralai-Mistral-7B-Instruct-v0-2

The Mistral7BInstructv0.2 Large Language Model (LLM) is an instruct finetuned version of the Mistral7Bv0.2. Mistral7Bv0.2 has the following changes compared to Mistral7Bv0.1: 32k context window (vs 8k context in v0.1) Ropetheta = 1e6 No SlidingWindow Attention For full details of

chat-completion
mistralai-Mixtral-8x7B-v01
mistralai-Mixtral-8x7B-v01

Model Details The Mixtral8x7Bv0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mixtral8x7Bv0.1 outperforms Llama 2 70B on most benchmarks with 6x faster inference. For full details of this model please read [release blog post](https://mistr

text-generation
mistral-community-Mixtral-8x22B-v0-1
mistral-community-Mixtral-8x22B-v0-1

The Mixtral8x22B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. Mixtral8x22Bv0.1 is a pretrained base model and therefore does not have any moderation mechanisms. Evaluation Results [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/Hugg

text-generation
Mistral-Large-2411
Mistral-Large-2411

Mistral Large 24.11 offers enhanced system prompts, advanced reasoning and function calling capabilities.

chat-completion